Automatic Sample-by-sample Model Selection Between Two Off-the-shelf Classifiers

نویسنده

  • Steve P. Chadwick
چکیده

If one could predict which of two classifiers will correctly classify a particular sample, then one could use the better classifier. Continuing this selection process throughout the data set should result in improved accuracy over either classifier alone. Fortunately, scalar measures which relate to the degree of confidence that we have in a classification can be computed for most common classifiers (Hastie & Tibshirani 1996). Some examples of confidence measures are distance from a linear discriminant separating plane (Duda & Hart 1973), distance to the nearest neighbor, distance to the nearest unlike neighbor, and distance to the center of correctly classified training data. We propose to apply discriminant analysis to the confidence measures, producing a rule which determines when one classifier is expected to be more accurate than the other. Let q1(x) and q2(x) be scalar functions for the confidence measures of two off-the-shelf classifiers. Each sample, xi, is mapped to (q1(xi), q2(xi)) in the decision space for selecting a classifier, thus the decision space has only two dimensions. Observe that the sample space has d-dimensions where d is the number of features in the sample. In this respect the dimensionality of selecting the classifier is reduced from d to 2. In order to select the better classifier, we need an estimate of where each classifier succeeds or fails. Both classifiers are applied to each training sample to create this estimate. Classifiers which never misclassify a training sample, such as nearest neighbors, are evaluated by leave-one-out runs. Each training sample now has two confidence values, one from each confidence function. It is also known whether each classifier has correctly classified each of the training samples. This classification information is used to associate a value selected from {−1, 0, 1} with each training sample. This value is termed correctness.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Cut-off Sampling Design: Take all, Take Some, and Take None

Extended Abstract. Sampling is the process of selecting units (e.g., people, organizations) from a population of interest so that by studying the sample we may fairly generalize our results back to the population from which they were chosen. To draw a sample from the underlying population, a variety of sampling methods can be employed, individually or in combination. Cut-off sampling is a pr...

متن کامل

Model Selection for Mixture Models Using Perfect Sample

We have considered a perfect sample method for model selection of finite mixture models with either known (fixed) or unknown number of components which can be applied in the most general setting with assumptions on the relation between the rival models and the true distribution. It is, both, one or neither to be well-specified or mis-specified, they may be nested or non-nested. We consider mixt...

متن کامل

Evaluation of Classifiers in Software Fault-Proneness Prediction

Reliability of software counts on its fault-prone modules. This means that the less software consists of fault-prone units the more we may trust it. Therefore, if we are able to predict the number of fault-prone modules of software, it will be possible to judge the software reliability. In predicting software fault-prone modules, one of the contributing features is software metric by which one ...

متن کامل

Generalization error for multi-class margin classification

In this article, we study rates of convergence of the generalization error of multi-class margin classifiers. In particular, we develop an upper bound theory quantifying the generalization error of various large margin classifiers. The theory permits a treatment of general margin losses, convex or nonconvex, in presence or absence of a dominating class. Three main results are established. First...

متن کامل

Exploiting Diversity for Natural Language Parsing

The popularity of applying machine learning methods to computational linguistics problems has produced a large supply of trainable natural language processing systems. Most problems of interest have an array of off-the-shelf products or downloadable code implementing solutions using various techniques. Where these solutions are developed independently, it is observed that their errors tend to b...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1999